Appendix Fugue: Slow-Worker-Agnostic Distributed Learning for Big Models on Big Data

ثبت نشده
چکیده

Table 1: Final tuned parameter values for Fugue, BarrieredFugue, GraphLab and PSGD. All the methods are tuned to perform optimally. η0 is the initial step size, where η is defined in equation 4. λ is the Dictionary Learning `1 penalty defined in equation 2. η ′ is parameter that modifies the learning rate when extra updates are executed while waiting for slow workers. step dec is a parameter for decreasing learning rate for GraphLab used in their collaborative filtering library.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Fugue: Slow-Worker-Agnostic Distributed Learning for Big Models on Big Data

We present a scheme for fast, distributed learning on big (i.e. high-dimensional) models applied to big datasets. Unlike algorithms that focus on distributed learning in either the big data or big model setting (but not both), our scheme partitions both the data and model variables simultaneously. This not only leads to faster learning on distributed clusters, but also enables machine learning ...

متن کامل

Distributed Structured Prediction for Big Data

The biggest limitations of learning structured predictors from big data are the computation time and the memory demands. In this paper, we propose to handle those big data problems efficiently by distributing and parallelizing the resource requirements. We present a distributed structured prediction learning algorithm for large scale models that cannot be effectively handled by a single cluster...

متن کامل

Study of Machine Learning Methods in Intelligent Transportation Systems

......................................................................................................................................... iii ACKNOWLEDGEMENTS .............................................................................................................. v LIST OF TABLES .................................................................................................................

متن کامل

Jubatus: An Open Source Platform for Distributed Online Machine Learning

Distributed computing is essential for handling very large datasets. Online learning is also promising for learning from rapid data streams. However, it is still an unresolved problem how to combine them for scalable learning and prediction on big data streams. We propose a general computational framework called loose model sharing for online and distributed machine learning. The key is to shar...

متن کامل

Big Learning with Bayesian Methods

The explosive growth in data volume and the availability of cheap computing resources have sparked increasing interest in Big learning, an emerging subfield that studies scalable machine learning algorithms, systems and applications with Big Data. Bayesian methods represent one important class of statistical methods for machine learning, with substantial recent developments on adaptive, flexibl...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2014